Mirror Descent Meets Fixed Share (and feels no regret)
نویسندگان
چکیده
Mirror descent with an entropic regularizer is known to achieve shifting regretbounds that are logarithmic in the dimension. This is done using either a carefullydesigned projection or by a weight sharing technique. Via a novel unified analysis,we show that these two approaches deliver essentially equivalent bounds on a no-tion of regret generalizing shifting, adaptive, discounted, and other related regrets.Our analysis also captures and extends the generalized weight sharing techniqueof Bousquet and Warmuth, and can be refined in several ways, including improve-ments for small losses and adaptive tuning of parameters.
منابع مشابه
A survey of Algorithms and Analysis for Adaptive Online Learning
We present tools for the analysis of Follow-The-Regularized-Leader (FTRL), Dual Averaging, and Mirror Descent algorithms when the regularizer (equivalently, proxfunction or learning rate schedule) is chosen adaptively based on the data. Adaptivity can be used to prove regret bounds that hold on every round, and also allows for data-dependent regret bounds as in AdaGrad-style algorithms (e.g., O...
متن کاملAnalysis Techniques for Adaptive Online Learning
We present tools for the analysis of Follow-The-Regularized-Leader (FTRL), Dual Averaging, and Mirror Descent algorithms when the regularizer (equivalently, proxfunction or learning rate schedule) is chosen adaptively based on the data. Adaptivity can be used to prove regret bounds that hold on every round, and also allows for data-dependent regret bounds as in AdaGrad-style algorithms (e.g., O...
متن کاملOn the Universality of Online Mirror Descent
We show that for a general class of convex online learning problems, Mirror Descent can always achieve a (nearly) optimal regret guarantee.
متن کاملSequential Change-Point Detection via Online Convex Optimization
Sequential change-point detection when the distribution parameters are unknown is a fundamental problem in statistics and machine learning. When the post-change parameters are unknown, we consider a set of detection procedures based on sequential likelihood ratios with non-anticipating estimators constructed using online convex optimization algorithms such as online mirror descent, which provid...
متن کاملOn near optimality of one-sample update for joint detection and estimation
Sequential hypothesis test and change-point detection when the distribution parameters 1 are unknown is a fundamental problem in statistics and machine learning. We show that for such 2 problems, detection procedures based on sequential likelihood ratios with simple one-sample update 3 estimates such as online mirror descent are nearly second-order optimal. This means that the upper 4 bound for...
متن کامل